Goto

Collaborating Authors

 unsupervised learning approach


Data driven approach towards more efficient Newton-Raphson power flow calculation for distribution grids

Yan, Shengyuan, Vazinram, Farzad, Kaseb, Zeynab, Spoor, Lindsay, Stiasny, Jochen, Mamudi, Betul, Ardakani, Amirhossein Heydarian, Orji, Ugochukwu, Vergara, Pedro P., Xiang, Yu, Guo, Jerry

arXiv.org Artificial Intelligence

Power flow (PF) calculations are fundamental to power system analysis to ensure stable and reliable grid operation. The Newton-Raphson (NR) method is commonly used for PF analysis due to its rapid convergence when initialized properly. However, as power grids operate closer to their capacity limits, ill-conditioned cases and convergence issues pose significant challenges. This work, therefore, addresses these challenges by proposing strategies to improve NR initialization, hence minimizing iterations and avoiding divergence. We explore three approaches: (i) an analytical method that estimates the basin of attraction using mathematical bounds on voltages, (ii) Two data-driven models leveraging supervised learning or physics-informed neural networks (PINNs) to predict optimal initial guesses, and (iii) a reinforcement learning (RL) approach that incrementally adjusts voltages to accelerate convergence. These methods are tested on benchmark systems. This research is particularly relevant for modern power systems, where high penetration of renewables and decentralized generation require robust and scalable PF solutions. In experiments, all three proposed methods demonstrate a strong ability to provide an initial guess for Newton-Raphson method to converge with fewer steps. The findings provide a pathway for more efficient real-time grid operations, which, in turn, support the transition toward smarter and more resilient electricity networks.


Unsupervised Learning Approach to Anomaly Detection in Gravitational Wave Data

Fayad, Ammar

arXiv.org Artificial Intelligence

Gravitational waves (GW), predicted by Einstein's General Theory of Relativity, provide a powerful probe of astrophysical phenomena and fundamental physics. In this work, we propose an unsupervised anomaly detection method using variational autoencoders (VAEs) to analyze GW time-series data. By training on noise-only data, the VAE accurately reconstructs noise inputs while failing to reconstruct anomalies, such as GW signals, which results in measurable spikes in the reconstruction error. The method was applied to data from the LIGO H1 and L1 detectors. Evaluation on testing datasets containing both noise and GW events demonstrated reliable detection, achieving an area under the ROC curve (AUC) of 0.89. This study introduces VAEs as a robust, unsupervised approach for identifying anomalies in GW data, which offers a scalable framework for detecting known and potentially new phenomena in physics.


Unsupervised Learning Approaches for Identifying ICU Patient Subgroups: Do Results Generalise?

Mayne, Harry, Parsons, Guy, Mahdi, Adam

arXiv.org Artificial Intelligence

The use of unsupervised learning to identify patient subgroups has emerged as a potentially promising direction to improve the efficiency of Intensive Care Units (ICUs). By identifying subgroups of patients with similar levels of medical resource need, ICUs could be restructured into a collection of smaller subunits, each catering to a specific group. However, it is unclear whether common patient subgroups exist across different ICUs, which would determine whether ICU restructuring could be operationalised in a standardised manner. In this paper, we tested the hypothesis that common ICU patient subgroups exist by examining whether the results from one existing study generalise to a different dataset. We extracted 16 features representing medical resource need and used consensus clustering to derive patient subgroups, replicating the previous study. We found limited similarities between our results and those of the previous study, providing evidence against the hypothesis. Our findings imply that there is significant variation between ICUs; thus, a standardised restructuring approach is unlikely to be appropriate. Instead, potential efficiency gains might be greater when the number and nature of the subunits are tailored to each ICU individually.


Supervised & Unsupervised Approach to Topic Modelling in Python

#artificialintelligence

This article will provide a high level intuition behind topic modelling and its associated applications. It will do a deep dive into various ways one can approach solving a problem which requires topic modelling and how you can solve those problems in both a supervised and unsupervised manner. I placed an emphasis on restructuring the data and initial problem such that the solution can be executed in a variety of methods. Topic modelling is a subsection of natural language processing (NLP) or text mining which aims to build models in order to parse various bodies of text with the goal of identifying topics mapped to the text. These models assist in identifying big picture topics associated with documents at scale.


GANs, Let's Give it a Deep Dive

#artificialintelligence

There are many ways a machine can be taught to generate an output on unseen data. The technological advancement in different sectors has left everyone shocked. The technique is none other than GAN(Generative Adversarial Network), which we're about to explore! But just before we begin -- here's the Jupyter Notebook with the code. Generative Adversarial Networks or GANs for short were developed in 2014 by Ian Goodfellow and his teammates.


Graph Learning for Fake Review Detection

#artificialintelligence

Fake reviews have become prevalent on various social networks such as e-commerce and social media platforms. As fake reviews cause a heavily negative influence on the public, timely detection and response are of great significance. To this end, effective fake review detection has become an emerging research area that attracts increasing attention from various disciplines like network science, computational social science, and data science. An important line of research in fake review detection is to utilize graph learning methods, which incorporate both the attribute features of reviews and their relationships into the detection process. To further compare these graph learning methods in this paper, we conduct a detailed survey on fake review detection. The survey presents a comprehensive taxonomy and covers advancements in three high-level categories, including fake review detection, fake reviewer detection, and fake review analysis. Different kinds of fake reviews and their corresponding examples are also summarized. Furthermore, we discuss the graph learning methods, including supervised and unsupervised learning approaches for fake review detection. Specifically, we outline the unsupervised learning approach that includes generation-based and contrast-based methods, respectively. In view of the existing problems in the current methods and data, we further discuss some challenges and open issues in this field, including the imperfect data, explainability, model efficien...


Unsupervised Feature Learning for Event Data: Direct vs Inverse Problem Formulation

Kostadinov, Dimche, Scaramuzza, Davide

arXiv.org Artificial Intelligence

Event-based cameras record an asynchronous stream of per-pixel brightness changes. As such, they have numerous advantages over the standard frame-based cameras, including high temporal resolution, high dynamic range, and no motion blur. Due to the asynchronous nature, efficient learning of compact representation for event data is challenging. While it remains not explored the extent to which the spatial and temporal event "information" is useful for pattern recognition tasks. In this paper, we focus on single-layer architectures. We analyze the performance of two general problem formulations: the direct and the inverse, for unsupervised feature learning from local event data (local volumes of events described in space-time). We identify and show the main advantages of each approach. Theoretically, we analyze guarantees for an optimal solution, possibility for asynchronous, parallel parameter update, and the computational complexity. We present numerical experiments for object recognition. We evaluate the solution under the direct and the inverse problem and give a comparison with the state-of-the-art methods. Our empirical results highlight the advantages of both approaches for representation learning from event data. We show improvements of up to 9 % in the recognition accuracy compared to the state-of-the-art methods from the same class of methods.


An unsupervised learning approach to solving heat equations on chip based on Auto Encoder and Image Gradient

He, Haiyang, Pathak, Jay

arXiv.org Machine Learning

Solving heat transfer equations on chip becomes very critical in the upcoming 5G and AI chip-package-systems. However, batches of simulations have to be performed for data driven supervised machine learning models. Data driven methods are data hungry, to address this, Physics Informed Neural Networks (PINN) have been proposed. However, vanilla PINN models solve one fixed heat equation at a time, so the models have to be retrained for heat equations with different source terms. Additionally, issues related to multi-objective optimization have to be resolved while using PINN to minimize the PDE residual, satisfy boundary conditions and fit the observed data etc. Therefore, this paper investigates an unsupervised learning approach for solving heat transfer equations on chip without using solution data and generalizing the trained network for predicting solutions for heat equations with unseen source terms. Specifically, a hybrid framework of Auto Encoder (AE) and Image Gradient (IG) based network is designed. The AE is used to encode different source terms of the heat equations. The IG based network implements a second order central difference algorithm for structured grids and minimizes the PDE residual. The effectiveness of the designed network is evaluated by solving heat equations for various use cases. It is proved that with limited number of source terms to train the AE network, the framework can not only solve the given heat transfer problems with a single training process, but also make reasonable predictions for unseen cases (heat equations with new source terms) without retraining.


Helm.ai raises $13M on its unsupervised learning approach to driverless car AI – TechCrunch

#artificialintelligence

Four years ago, mathematician Vlad Voroninski saw an opportunity to remove some of the bottlenecks in the development of autonomous vehicle technology thanks to breakthroughs in deep learning. Now, Helm.ai, the startup he co-founded in 2016 with Tudor Achim, is coming out of stealth with an announcement that it has raised $13 million in a seed round that includes investment from A.Capital Ventures, Amplo, Binnacle Partners, Sound Ventures, Fontinalis Partners and SV Angel. More than a dozen angel investors also participated, including Berggruen Holdings founder Nicolas Berggruen, Quora co-founders Charlie Cheever and Adam D'Angelo, professional NBA player Kevin Durant, Gen. David Petraeus, Matician co-founder and CEO Navneet Dalal, Quiet Capital managing partner Lee Linden and Robinhood co-founder Vladimir Tenev, among others. Helm.ai will put the $13 million in seed funding toward advanced engineering and R&D and hiring more employees, as well as locking in and fulfilling deals with customers. Helm.ai is focused solely on the software.


Detecting abnormalities in resting-state dynamics: An unsupervised learning approach

Khosla, Meenakshi, Jamison, Keith, Kuceyeski, Amy, Sabuncu, Mert R.

arXiv.org Machine Learning

Much of the research in this direction has aimed at identifying connectivity based biomarkers, restricting the analysis to so-called "static" functional connectivity measures that quantify the average degree of synchrony between brain regions. For e.g., machine learning based strategies have been used with static connectivity measures to parcellate the brain into functional networks, and extract individual-level predictions about cognitive state or clinical condition [2]. In recent years, there has been a surge in the study of the temporal dynamics of rsfMRI data, offering a complementary perspective on the functional connectome and how it is altered in disease, development, and aging [14]. However, to our knowledge, there has been a dearth of machine learning applications to dynamic rsfMRI analysis. Thanks to large-scale datasets, modern machine learning methods have fueled significant progress in computer vision. Compared to natural vision applications, however, medical imaging poses a unique set of challenges. Data, particularly labeled data, are often scarce in medical imaging applications. This makes data-hungry methods such as supervised CNNs possibly less useful. One potential approach to tackle the limited sample size issue is to exploit unsupervised arXiv:1908.06168v1